VQ VAE


Vector-quantized variational autoencoder (VQ VAE) is a generative model that uses vector quantization to learn discrete latent representations.

Modality-Aware and Anatomical Vector-Quantized Autoencoding for Multimodal Brain MRI

Add code
Apr 06, 2026
Viaarxiv icon

Collapse-Free Prototype Readout Layer for Transformer Encoders

Add code
Apr 04, 2026
Viaarxiv icon

Investigating Permutation-Invariant Discrete Representation Learning for Spatially Aligned Images

Add code
Apr 02, 2026
Viaarxiv icon

MAESIL: Masked Autoencoder for Enhanced Self-supervised Medical Image Learning

Add code
Apr 01, 2026
Viaarxiv icon

AceTone: Bridging Words and Colors for Conditional Image Grading

Add code
Apr 01, 2026
Viaarxiv icon

BAT: Balancing Agility and Stability via Online Policy Switching for Long-Horizon Whole-Body Humanoid Control

Add code
Apr 01, 2026
Viaarxiv icon

CardioDiT: Latent Diffusion Transformers for 4D Cardiac MRI Synthesis

Add code
Mar 26, 2026
Viaarxiv icon

SIMART: Decomposing Monolithic Meshes into Sim-ready Articulated Assets via MLLM

Add code
Mar 24, 2026
Viaarxiv icon

Perceptio: Perception Enhanced Vision Language Models via Spatial Token Generation

Add code
Mar 19, 2026
Viaarxiv icon

Architecture-Agnostic Feature Synergy for Universal Defense Against Heterogeneous Generative Threats

Add code
Mar 16, 2026
Viaarxiv icon